reference trajectory
Appendix 571 In this appendix, we provide more details about the four experiments and some scenario examples
Autoregressive sampling is used to create a traffic snapshot. We train a scenario generation model TrafficGen with mixed data. The detailed hyperparameters are shown in Table 4. Figure 7: Dynamics of the generated traffic scenarios. The first column is the original case. The middle columns show the generated scenarios at different timesteps.
- Energy (0.47)
- Transportation (0.30)
A Problem Formulation using L1 and L
Proof of Lemma 2. Let U be the data set associated to ν. Proof of Lemma 3. First, we prove that the property holds for the root node. We wish to prove the property for some unexplored leaf after the iteration. This is trivial if the leaf ν is not expanded in that iteration. Suppose the leaf ν is expanded. Proof of Lemma 5. From Lemma 2, we note that Q Consider any path from the root to a leaf whose length is mK for some integer K > 0. We note that for each node ν and any of its children ν (Lemma 5).
MARINE: Theoretical Optimization and Design for Multi-Agent Recursive IN-context Enhancement
Zhang, Hongwei, Lu, Ji, Du, Yongsheng, Gao, Yanqin, Huang, Lingjun, Wang, Baoli, Tan, Fang, Zou, Peng
Large Language Model (LLM)-based agents demonstrate advanced reasoning capabilities, yet practical constraints frequently limit outputs to single responses, leaving significant performance potential unrealized. This paper introduces MARINE (Multi-Agent Recursive IN-context Enhancement), a theoretically grounded framework that reconceptualizes test-time reasoning as iterative refinement of a persistent reference trajectory, fundamentally departing from conventional one-shot or multi-sample paradigms. The MARINE refinement operator systematically converts a base model's pass@N capabilities into near-optimal pass@1 performance. Rigorous theoretical analysis establishes that minimal feasible batches maximize expected performance gains under fixed invocation budgets, while logarithmically growing batch schedules ensure continuous improvement without computational constraints. Comprehensive evaluation on the BrowserComp-ZH benchmark demonstrates state-of-the-art results, with a 685B-parameter implementation achieving 46.0% pass@1 accuracy. Meanwhile, MARINE establishes a new paradigm for parameter-efficient reasoning: an 80B-parameter model augmented with MARINE matches the performance of standalone 1000B-parameter agents, reducing parameter requirements by over an order of magnitude. Notably, within a fixed computational budget, the proposed MARINE delivers higher-quality samples to alignment and optimization processes than traditional sampling-and-ranking strategies. Consequently, it has great potential to boost post-training efficiency.
- Europe > Switzerland > Basel-City > Basel (0.04)
- Asia > China (0.04)
- Research Report (0.64)
- Overview (0.47)
UMI-on-Air: Embodiment-Aware Guidance for Embodiment-Agnostic Visuomotor Policies
Gupta, Harsh, Guo, Xiaofeng, Ha, Huy, Pan, Chuer, Cao, Muqing, Lee, Dongjae, Scherer, Sebastian, Song, Shuran, Shi, Guanya
We introduce UMI-on-Air, a framework for embodiment-aware deployment of embodiment-agnostic manipulation policies. Our approach leverages diverse, unconstrained human demonstrations collected with a handheld gripper (UMI) to train generalizable visuomotor policies. A central challenge in transferring these policies to constrained robotic embodiments-such as aerial manipulators-is the mismatch in control and robot dynamics, which often leads to out-of-distribution behaviors and poor execution. To address this, we propose Embodiment-Aware Diffusion Policy (EADP), which couples a high-level UMI policy with a low-level embodiment-specific controller at inference time. By integrating gradient feedback from the controller's tracking cost into the diffusion sampling process, our method steers trajectory generation towards dynamically feasible modes tailored to the deployment embodiment. This enables plug-and-play, embodiment-aware trajectory adaptation at test time. We validate our approach on multiple long-horizon and high-precision aerial manipulation tasks, showing improved success rates, efficiency, and robustness under disturbances compared to unguided diffusion baselines. Finally, we demonstrate deployment in previously unseen environments, using UMI demonstrations collected in the wild, highlighting a practical pathway for scaling generalizable manipulation skills across diverse-and even highly constrained-embodiments. All code, data, and checkpoints will be publicly released after acceptance. Result videos can be found at umi-on-air.github.io.
Model-Less Feedback Control of Space-based Continuum Manipulators using Backbone Tension Optimization
Rajneesh, Shrreya, Pavle, Nikita, Sahoo, Rakesh Kumar, Sinha, Manoranjan
Continuum manipulators offer intrinsic dexterity and safe geometric compliance for navigation within confined and obstacle-rich environments. However, their infinite-dimensional backbone deformation, unmodeled internal friction, and configuration-dependent stiffness fundamentally limit the reliability of model-based kinematic formulations, resulting in inaccurate Jacobian predictions, artificial singularities, and unstable actuation behavior. Motivated by these limitations, this work presents a complete model-less control framework that bypasses kinematic modeling by using an empirically initialized Jacobian refined online through differential convex updates. Tip motion is generated via a real-time quadratic program that computes actuator increments while enforcing tendon slack avoidance and geometric limits. A backbone-tension optimization term is introduced in this paper to regulate axial loading and suppress co-activation compression. The framework is validated across circular, pentagonal, and square trajectories, demonstrating smooth convergence, stable tension evolution, and sub-millimeter steady-state accuracy without any model calibration or parameter identification. These results establish the proposed controller as a scalable alternative to model-dependent continuum manipulation in a constrained environment.
- Asia > India > West Bengal > Kharagpur (0.05)
- North America > United States (0.04)
Sliding Mode Control and Subspace Stabilization Methodology for the Orbital Stabilization of Periodic Trajectories
Surov, Maksim, Freidovich, Leonid
The problem of orbital stabilization of periodic trajectories has been addressed in a series of publications: [1, 2, 3, 4, 5, 6, 7]. Many of these works, e.g., [1, 2, 4, 7], employ the transverse linearization approach, which approximates the dynamics near a reference periodic orbit by a linear time-varying (LTV) system with periodic coefficients. As shown in [2, 8], a feedback designed to stabilize the trivial solution of this auxiliary LTV system can be used to construct a control law that stabilizes the orbit of the original nonlinear system. Under the mild assumption of controllability of the LTV system over one period, the LQR approach can be used to design the feedback. The practical effectiveness of this method was demonstrated in experiments with real robotic systems in [9, 10, 11]. A substantially different stabilization method for the LTV system was proposed in [5], where the authors developed an alternative scheme combining Floquet theory with sliding-mode control. Following this line of work, we show that a specific feedback linearization of the transverse dynamics yields an LTV system endowed with a stable invariant subspace. In this setting, the control objective reduces to driving all trajectories into the stable subspace, which is achieved via sliding-mode-based control. This method does not require solving the computationally demanding periodic LQR problem.
- Europe > Austria > Vienna (0.14)
- North America > United States > New York (0.04)
- North America > United States > Colorado > Denver County > Denver (0.04)
- (4 more...)
On Disturbance-Aware Minimum-Time Trajectory Planning: Evidence from Tests on a Dynamic Driving Simulator
Masoni, Matteo, Palermo, Vincenzo, Gabiccini, Marco, Gulisano, Martino, Previati, Giorgio, Gobbi, Massimiliano, Comolli, Francesco, Mastinu, Gianpiero, Guiggiani, Massimo
This work investigates how disturbance-aware, robustness-embedded reference trajectories translate into driving performance when executed by professional drivers in a dynamic simulator. Three planned reference trajectories are compared against a free-driving baseline (NOREF) to assess trade-offs between lap time (LT) and steering effort (SE): NOM, the nominal time-optimal trajectory; TLC, a track-limit-robust trajectory obtained by tightening margins to the track edges; and FLC, a friction-limit-robust trajectory obtained by tightening against axle and tire saturation. All trajectories share the same minimum lap-time objective with a small steering-smoothness regularizer and are evaluated by two professional drivers using a high-performance car on a virtual track. The trajectories derive from a disturbance-aware minimum-lap-time framework recently proposed by the authors, where worst-case disturbance growth is propagated over a finite horizon and used to tighten tire-friction and track-limit constraints, preserving performance while providing probabilistic safety margins. LT and SE are used as performance indicators, while RMS lateral deviation, speed error, and drift angle characterize driving style. Results show a Pareto-like LT-SE trade-off: NOM yields the shortest LT but highest SE; TLC minimizes SE at the cost of longer LT; FLC lies near the efficient frontier, substantially reducing SE relative to NOM with only a small LT increase. Removing trajectory guidance (NOREF) increases both LT and SE, confirming that reference trajectories improve pace and control efficiency. Overall, the findings highlight reference-based and disturbance-aware planning, especially FLC, as effective tools for training and for achieving fast yet stable trajectories.
- Europe > Italy > Tuscany > Pisa Province > Pisa (0.04)
- Europe > Italy > Lombardy > Milan (0.04)
- Europe > Germany > North Rhine-Westphalia > Cologne Region > Aachen (0.04)
- Transportation > Ground > Road (1.00)
- Leisure & Entertainment > Sports > Motorsports (1.00)
- Automobiles & Trucks (1.00)
ConfRover: Simultaneous Modeling of Protein Conformation and Dynamics via Autoregression
Shen, Yuning, Wang, Lihao, Yuan, Huizhuo, Wang, Yan, Yang, Bangji, Gu, Quanquan
Understanding protein dynamics is critical for elucidating their biological functions. The increasing availability of molecular dynamics (MD) data enables the training of deep generative models to efficiently explore the conformational space of proteins. However, existing approaches either fail to explicitly capture the temporal dependencies between conformations or do not support direct generation of time-independent samples. To address these limitations, we introduce ConfRover, an autoregressive model that simultaneously learns protein conformation and dynamics from MD trajectories, supporting both time-dependent and time-independent sampling. At the core of our model is a modular architecture comprising: (i) an encoding layer, adapted from protein folding models, that embeds protein-specific information and conformation at each time frame into a latent space; (ii) a temporal module, a sequence model that captures conformational dynamics across frames; and (iii) an SE(3) diffusion model as the structure decoder, generating conformations in continuous space. Experiments on ATLAS, a large-scale protein MD dataset of diverse structures, demonstrate the effectiveness of our model in learning conformational dynamics and supporting a wide range of downstream tasks. ConfRover is the first model to sample both protein conformations and trajectories within a single framework, offering a novel and flexible approach for learning from protein MD data. Project website: https://bytedance-seed.github.io/ConfRover.
MIMIC-MJX: Neuromechanical Emulation of Animal Behavior
Zhang, Charles Y., Yang, Yuanjia, Sirbu, Aidan, Abe, Elliott T. T., Wärnberg, Emil, Leonardis, Eric J., Aldarondo, Diego E., Lee, Adam, Prasad, Aaditya, Foat, Jason, Bian, Kaiwen, Park, Joshua, Bhatt, Rusham, Saunders, Hutton, Nagamori, Akira, Thanawalla, Ayesha R., Huang, Kee Wui, Plum, Fabian, Beck, Hendrik K., Flavell, Steven W., Labonte, David, Richards, Blake A., Brunton, Bingni W., Azim, Eiman, Ölveczky, Bence P., Pereira, Talmo D.
The primary output of the nervous system is movement and behavior. While recent advances have democratized pose tracking during complex behavior, kinematic trajectories alone provide only indirect access to the underlying control processes. Here we present MIMIC-MJX, a framework for learning biologically-plausible neural control policies from kinematics. MIMIC-MJX models the generative process of motor control by training neural controllers that learn to actuate biomechanically-realistic body models in physics simulation to reproduce real kinematic trajectories. We demonstrate that our implementation is accurate, fast, data-efficient, and generalizable to diverse animal body models. Policies trained with MIMIC-MJX can be utilized to both analyze neural control strategies and simulate behavioral experiments, illustrating its potential as an integrative modeling framework for neuroscience.
- North America > United States > Washington > King County > Seattle (0.14)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > California > San Diego County > La Jolla (0.04)
- (7 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Robots (0.93)
Switching-time bioprocess control with pulse-width-modulated optogenetics
Biotechnology can benefit from dynamic control to improve production efficiency. In this context, optogenetics enables modulation of gene expression using light as an external input, allowing fine-tuning of protein levels to unlock dynamic metabolic control and regulation of cell growth. Optogenetic systems can be actuated by light intensity. However, relying solely on intensity-driven control (i.e., signal amplitude) may fail to properly tune optogenetic bioprocesses when the dose-response relationship (i.e., light intensity versus gene-expression strength) is steep. In these cases, tunability is effectively constrained to either fully active or fully repressed gene expression, with little intermediate regulation. Pulse-width modulation, a concept widely used in electronics, can alleviate this issue by alternating between fully ON and OFF light intensity within forcing periods, thereby smoothing the average response and enhancing process controllability. Naturally, optimizing pulse-width-modulated optogenetics entails a switching-time optimal control problem with a binary input over many forcing periods. While this can be formulated as a mixed-integer program on a refined time grid, the number of decision variables can grow rapidly with increasing time-grid resolution and number of forcing periods, compromising tractability. Here, we propose an alternative solution based on reinforcement learning. We parametrize control actions via the duty cycle, a continuous variable that encodes the ON-to-OFF switching time within each forcing period, thereby respecting the intrinsic binary nature of the light intensity.
- North America > United States (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)